2 research outputs found

    Searching for a partially absorbing target by a run-and-tumble particle in a confined space

    Full text link
    A random search of a partially absorbing target by a run-and-tumble particle in a confined one-dimensional space is investigated. We analytically obtain the mean searching time, which shows a non-monotonic behavior as a function of the self-propulsion speed of the active particle, indicating the existence of an optimal speed, when the absorption strength of the target is finite. In the limit of large and small absorption strengths, respectively, asymptotes of the mean searching time and the optimal speed are found. We also demonstrate that the first-passage problem of a diffusive run-and-tumble particle in high dimensions can be mapped into a one-dimensional problem with a partially absorbing target. Finally, as a practical application exploiting the existence of the optimal speed, we propose a filtering device to extract active particles with a desired speed and evaluate how the resolution of the filtering device depends on the absorption strength.Comment: 12 pages, 3 figure

    Tree of Clarifications: Answering Ambiguous Questions with Retrieval-Augmented Large Language Models

    Full text link
    Questions in open-domain question answering are often ambiguous, allowing multiple interpretations. One approach to handling them is to identify all possible interpretations of the ambiguous question (AQ) and to generate a long-form answer addressing them all, as suggested by Stelmakh et al., (2022). While it provides a comprehensive response without bothering the user for clarification, considering multiple dimensions of ambiguity and gathering corresponding knowledge remains a challenge. To cope with the challenge, we propose a novel framework, Tree of Clarifications (ToC): It recursively constructs a tree of disambiguations for the AQ -- via few-shot prompting leveraging external knowledge -- and uses it to generate a long-form answer. ToC outperforms existing baselines on ASQA in a few-shot setup across the metrics, while surpassing fully-supervised baselines trained on the whole training set in terms of Disambig-F1 and Disambig-ROUGE. Code is available at https://github.com/gankim/tree-of-clarifications.Comment: Accepted to EMNLP 202
    corecore